# High-precision visual representation
Eva02 Base Patch16 Clip 224.merged2b
MIT
The EVA CLIP model is a vision-language model built on the OpenCLIP and timm frameworks, supporting tasks like zero-shot image classification.
Text-to-Image
E
timm
3,029
0
Eva02 Large Patch14 Clip 224.merged2b S4b B131k
MIT
EVA02 is a large-scale vision-language model based on the CLIP architecture, supporting zero-shot image classification tasks.
Image Classification
E
timm
5,696
6
Featured Recommended AI Models